# Long-context processing

Greenmind Medium 14B R1
MIT
GreenMind-Medium-14B-R1 is a medium-scale Vietnamese language model capable of effectively solving problems requiring intermediate reasoning, such as common sense, mathematics, natural sciences, and social sciences topics.
Large Language Model Supports Multiple Languages
G
GreenNode
50
2
Kanana 1.5 2.1b Base
Apache-2.0
Kanana 1.5 is a new version of the Kanana model family, featuring significant improvements in coding, mathematics, and function calling capabilities, supporting 32K tokens length and extendable to 128K tokens.
Large Language Model Transformers Supports Multiple Languages
K
kakaocorp
757
4
EXAONE Deep 2.4B GGUF
Other
EXAONE Deep is an efficient reasoning language model developed by LG AI Research with 2.4B parameters, excelling in reasoning tasks such as mathematics and programming.
Large Language Model Supports Multiple Languages
E
Mungert
968
3
Olmo2 11B SuperBPE T180k
Apache-2.0
An 11-billion parameter large language model trained with the innovative SuperBPE tokenizer, supporting superword unit recognition and subword tokenization capabilities.
Large Language Model Transformers English
O
UW
29
2
Rumodernbert Base
Apache-2.0
A modern bidirectional encoder-only Transformer model for Russian, pre-trained on approximately 2 trillion Russian, English, and code tokens, with a context length of up to 8,192 tokens.
Large Language Model Transformers Supports Multiple Languages
R
deepvk
2,992
40
Llama 3 Korean Bllossom 70B
A Korean-English bilingual enhanced model based on Llama3, optimized for Korean language capabilities through vocabulary expansion and knowledge correlation
Large Language Model Transformers Supports Multiple Languages
L
Bllossom
300
86
Stripedhyena Nous 7B
Apache-2.0
StripedHyena-Nous-7B is a hybrid-architecture chat model combining multi-head grouped query attention and gated convolutions, supporting long-context processing.
Large Language Model Transformers English
S
togethercomputer
115
141
Stripedhyena Hessian 7B
Apache-2.0
StripedHyena-Hessian-7B is a Transformer alternative developed by Together Research based on a hybrid signal processing model, designed for long-context scenarios, offering more efficient training and inference performance.
Large Language Model Transformers English
S
togethercomputer
38
66
Mistrallite
Apache-2.0
MistralLite is a fine-tuned language model based on Mistral-7B-v0.1, enhanced for handling long contexts (up to 32K tokens), suitable for long-context retrieval, topic summarization, QA, and similar scenarios.
Large Language Model Transformers
M
AWS
61.78k
430
Chinese Llama 2 7b 16k
Apache-2.0
Chinese-LLaMA-2-7B-16K is a Chinese large language model developed based on Meta's Llama-2, supporting 16K context length, suitable for inference and full-parameter training.
Large Language Model Transformers Supports Multiple Languages
C
hfl
57
11
Llama 2 7b Chat Hf Function Calling V2
Llama 2 is a 7B-parameter dialogue-optimized large language model developed by Meta. This version extends function calling capabilities, supporting structured JSON format responses.
Large Language Model English
L
Trelis
175
136
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase